Incorporating A Rich Linguistic Model into Whole-Book Recognition

نویسندگان

  • Pingping Xiu
  • Henry S. Baird
چکیده

Whole-book recognition, a technique that improves recognition of book-images using fully automatic mutual-entropybased model adaptation, has achieved character error rate as low as 1.9% on 50 pages of real book images in our previous publications. However, the linguistic model for word recognition was simple, assuming a uniform distribution on the words in the dictionary, so that the algorithm is unaware of prior word-occurrence distribution. As a result, the statistics of the output transcript differs largely from that of a real distribution. In this paper, we propose a post-processing technique that improves the existing whole-book recognition results by applying the constraints of a rich linguistic model a prior word-occurrence distribution. This technique further drives the character error rate down from 1.9% to 0.97%. We also show that the wholebook recognition algorithm combined with this post-processing technique shows faster improvements in which word error rates fall monotonically with passage length.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Incorporating linguistic post-processing into whole-book recognition

We describe a technique of linguistic post-processing of whole-book recognition results. Whole-book recognition is a technique that improves recognition of book images using fully automatic cross-entropy-based model adaptation. In previous published works, word recognition was performed on individual words separately, without awaring passage-level information such as word-occurrence frequencies...

متن کامل

Whole-book recognition using mutual-entropy-driven model adaptation

We describe an approach to unsupervised high-accuracy recognition of the textual contents of an entire book using fully automatic mutual-entropy-based model adaptation. Given images of all the pages of a book together with approximate models of image formation (e.g. a character-image classifier) and linguistics (e.g. a word-occurrence probability model), we detect evidence for disagreements bet...

متن کامل

Incorporating linguistic knowledge and automatic baseform generation in acoustic subword unit based speech recognition

A major challenge in speech recognition based on acoustic subword units is creating a lexicon which is robust to interand intra-speaker variations. In this paper we present two di erent approaches for incorporating simple word-level linguistic knowledge into the labelling step of the training procedure. The proposed systems also utilise a scheme for combined optimisation of baseforms and subwor...

متن کامل

Incorporating Cognitive Linguistic Insights into Classrooms: the Case of Iranian Learners’ Acquisition of If-Clauses

Cognitive linguistics gives the most inclusive, consistent description of how language is organized, used and learned to date. Cognitive linguistics contains a great number of concepts that are useful to second language learners.  If-clauses in English, on the other hand, remain intriguing for foreign language learners to struggle with, due to their intrinsic intricacies. EFL grammar books are ...

متن کامل

Beyond N-Grams: Can Linguistic Sophistication Improve Language Modeling?

It seems obvious that a successful model of natural language would incorporate a great deal of both linguistic and world knowledge. Interestingly, state of the art language models for speech recognition are based on a very crude linguistic model, namely conditioning the probability of a word on a small fixed number of preceding words. Despite many attempts to incorporate more sophisticated info...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2009